Goto

Collaborating Authors

 reinforcement learning enhanced explainer


Reinforcement Learning Enhanced Explainer for Graph Neural Networks

Neural Information Processing Systems

Graph neural networks (GNNs) have recently emerged as revolutionary technologies for machine learning tasks on graphs. In GNNs, the graph structure is generally incorporated with node representation via the message passing scheme, making the explanation much more challenging. Given a trained GNN model, a GNN explainer aims to identify a most influential subgraph to interpret the prediction of an instance (e.g., a node or a graph), which is essentially a combinatorial optimization problem over graph. The existing works solve this problem by continuous relaxation or search-based heuristics. But they suffer from key issues such as violation of message passing and hand-crafted heuristics, leading to inferior interpretability.


Supplementary: Reinforcement Learning Enhanced Explainer for Graph Neural Networks Caihua Shan

Neural Information Processing Systems

(line 4). We show our RG-Explainer for graph classification in Alg. 2. The algorithm is similar to the one explaining node classifications, except that we train our seed locator to detect the most influential (line 4). Input: The input graph G = ( V, E), node features X, node instances I, and a trained GNN model f () . Check the stopping criteria by Eq. 10. I, and a trained GNN model f () .


Reinforcement Learning Enhanced Explainer for Graph Neural Networks

Neural Information Processing Systems

Graph neural networks (GNNs) have recently emerged as revolutionary technologies for machine learning tasks on graphs. In GNNs, the graph structure is generally incorporated with node representation via the message passing scheme, making the explanation much more challenging. Given a trained GNN model, a GNN explainer aims to identify a most influential subgraph to interpret the prediction of an instance (e.g., a node or a graph), which is essentially a combinatorial optimization problem over graph. The existing works solve this problem by continuous relaxation or search-based heuristics. But they suffer from key issues such as violation of message passing and hand-crafted heuristics, leading to inferior interpretability.